Overcoming Catastrophic Forgetting via Direction-Constrained Optimization

نویسندگان

چکیده

This paper studies a new design of the optimization algorithm for training deep learning models with fixed architecture classification network in continual framework. The data is non-stationary and non-stationarity imposed by sequence distinct tasks. We first analyze model trained on only one task isolation identify region parameter space, where performance close to recovered optimum. provide empirical evidence that this resembles cone expands along convergence direction. study principal directions trajectory optimizer after show traveling few top can quickly bring parameters outside but not case remaining directions. argue catastrophic forgetting setting be alleviated when are constrained stay within intersection plausible cones individual tasks were so far encountered during training. Based observation we present our direction-constrained (DCO) method, each introduce linear autoencoder approximate its corresponding forbidden They then incorporated into loss function form regularization term purpose coming without forgetting. Furthermore, order control memory growth as number increases, propose memory-efficient version called compressed DCO (DCO-COMP) allocates size storing all autoencoders. empirically demonstrate performs favorably compared other state-of-art regularization-based methods. codes publicly available at https://github.com/yunfei-teng/DCO .

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Overcoming Catastrophic Forgetting by Incremental Moment Matching

Catastrophic forgetting is a problem of neural networks that loses the information of the first task after training the second task. Here, we propose a method, i.e. incremental moment matching (IMM), to resolve this problem. IMM incrementally matches the moment of the posterior distribution of the neural network which is trained on the first and the second task, respectively. To make the search...

متن کامل

Overcoming catastrophic forgetting in neural networks

The ability to learn tasks in a sequential fashion is crucial to the development of artificial intelligence. Until now neural networks have not been capable of this and it has been widely thought that catastrophic forgetting is an inevitable feature of connectionist models. We show that it is possible to overcome this limitation and train networks that can maintain expertise on tasks that they ...

متن کامل

Overcoming catastrophic forgetting with hard attention to the task

Catastrophic forgetting occurs when a neural network loses the information learned in a previous task after training on subsequent tasks. This problem remains a hurdle for artificial intelligence systems with sequential learning capabilities. In this paper, we propose a task-based hard attention mechanism that preserves previous tasks’ information without affecting the current task’s learning. ...

متن کامل

Alleviating Catastrophic Forgetting via Multi-Objective Learning [IJCNN1762]

Handling catastrophic forgetting is an interesting and challenging topic in modeling the memory mechanisms of the human brain using machine learning models. From a more general point of view, catastrophic forgetting reflects the stability-plasticity dilemma, which is one of the several dilemmas to be addressed in learning systems: to retain the stored memory while learning new information. Diff...

متن کامل

Catastrophic forgetting in connectionist networks.

All natural cognitive systems, and, in particular, our own, gradually forget previously learned information. Plausible models of human cognition should therefore exhibit similar patterns of gradual forgetting of old information as new information is acquired. Only rarely does new learning in natural cognitive systems completely disrupt or erase previously learned information; that is, natural c...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Lecture Notes in Computer Science

سال: 2023

ISSN: ['1611-3349', '0302-9743']

DOI: https://doi.org/10.1007/978-3-031-26387-3_41